# Small Language Model

Doge 20M Chinese
Apache-2.0
The Doge model employs dynamic masked attention mechanisms for sequence transformation, with the option to use either multi-layer perceptrons or cross-domain mixture of experts for state transitions.
Large Language Model Transformers Supports Multiple Languages
D
wubingheng
65
2
Doge 320M Instruct
Apache-2.0
Doge 320M Instruct is a lightweight language model based on dynamic masked attention, trained with supervised fine-tuning (SFT) and direct preference optimization (DPO), suitable for question-answering and dialogue tasks.
Large Language Model Transformers English
D
SmallDoge
12.61k
3
Bonsai
Bonsai is a small ternary-weighted language model with 500 million parameters, built on the Llama architecture and using the Mistral tokenizer, trained on fewer than 5 billion tokens.
Large Language Model Transformers
B
deepgrove
113
8
Doge 160M
Apache-2.0
Doge 160M is a small language model that employs dynamic masked attention mechanisms, trained by the SmallDoge community, and supports text generation tasks.
Large Language Model Transformers Supports Multiple Languages
D
SmallDoge
4,227
4
Doge 20M Instruct
Apache-2.0
Doge 20M is a small language model based on dynamic masked attention mechanism, supporting instruction following and Q&A tasks.
Large Language Model Transformers English
D
SmallDoge
5,010
4
Simple Stories 4M
MIT
Simple Stories is a series of small text generation models trained on the TinyStories dataset, focusing on generating children's stories.
Text Generation Transformers English
S
broskicodes
104
16
Minilm L12 H384 Uncased Finetuned Imdb
MIT
This model is a fine-tuned version of Microsoft's MiniLM-L12-H384-uncased on the IMDB dataset, primarily used for sentiment analysis tasks.
Large Language Model Transformers
M
lewtun
18
0
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase